# Long Text Processing

Qwen3 235B A22B
Apache-2.0
Qwen3 is the latest generation of large language models in the Qwen series, offering a range of dense and Mixture of Experts (MoE) models. Based on extensive training, Qwen3 has achieved groundbreaking progress in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model Transformers
Q
unsloth
421
2
Led Financial Summarization Genai15
A fine-tuned LED model specifically for financial text summarization tasks, based on pszemraj/led-large-book-summary, developed by the Warwick Business School GEN AI GROUP 15 team
Text Generation Other
L
fahil2631
15
0
Qwen3 4B GGUF
Apache-2.0
Qwen3 is the latest version of the Tongyi Qianwen series of large language models, offering a range of dense and mixture-of-experts (MoE) models. Based on large-scale training, Qwen3 has achieved breakthrough progress in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model
Q
Qwen
4,225
6
Qwen3 14B 128K GGUF
Apache-2.0
Qwen3 is the latest generation of large language models in the Qwen series, offering a range of dense and mixture-of-experts (MoE) models. Based on extensive training, Qwen3 has achieved breakthrough progress in reasoning, instruction following, agent capabilities, and multilingual support.
Large Language Model English
Q
unsloth
10.20k
13
II Medical 7B Preview
A medical reasoning model fine-tuned based on Qwen/Qwen2.5-7B-Instruct, excelling in multiple medical QA benchmarks
Large Language Model Transformers
I
Intelligent-Internet
112
9
Qwen3 4B Unsloth Bnb 4bit
Apache-2.0
Qwen3-4B is the latest generation of the Qwen series large language model, featuring 4B parameters and supporting over 100 languages, with outstanding performance in reasoning, instruction following, and agent capabilities.
Large Language Model Transformers English
Q
unsloth
72.86k
5
Qwen3 4B GGUF
Apache-2.0
Qwen3-4B is the latest generation large language model in the Qwen series with 4B parameters, supporting over 100 languages and demonstrating exceptional reasoning, instruction following, and agent capabilities.
Large Language Model English
Q
unsloth
59.40k
32
Chonky Modernbert Large 1
MIT
Chonky is a Transformer model capable of intelligently splitting text into meaningful semantic chunks, suitable for RAG systems.
Sequence Labeling Transformers English
C
mirth
54
2
Llm Jp Modernbert Base
Apache-2.0
A Japanese large language model based on the modernBERT-base architecture, supporting a maximum sequence length of 8192, trained on 3.4TB of Japanese corpus
Large Language Model Transformers Japanese
L
llm-jp
1,398
5
Orpheus 3b 0.1 Ft.w8a8
Apache-2.0
Orpheus-3B-0.1-FT is a text-to-speech model based on a causal language model, supporting efficient quantization compression.
Large Language Model Transformers English
O
nytopop
173
0
Chonky Modernbert Base 1
MIT
Chonky is a Transformer model that intelligently splits text into meaningful semantic chunks for RAG systems.
Sequence Labeling Transformers English
C
mirth
221
1
Shisa V2 Unphi4 14b
Apache-2.0
Shisa V2 is a bilingual (Japanese/English) general-purpose chat model trained by Shisa.AI, focusing on enhancing Japanese task capabilities while maintaining strong English proficiency.
Large Language Model Transformers Supports Multiple Languages
S
shisa-ai
62
3
Ruri V3 Reranker 310m
Apache-2.0
A Japanese general-purpose reranking model built on ModernBERT-Ja, featuring top-tier performance and long-sequence processing capabilities
Text Embedding Japanese
R
cl-nagoya
1,100
5
Longformer Base 4096 Bible
Apache-2.0
A Bible text relevance prediction model fine-tuned based on allenai/longformer-base-4096
Text Classification Transformers English
L
ronenh24
32
1
Mizan Rerank V1
Apache-2.0
A revolutionary open-source model capable of reordering long Arabic texts with exceptional efficiency and accuracy.
Text Embedding Supports Multiple Languages
M
ALJIACHI
167
1
Reranker ModernBERT Base Gooaq 1 Epoch 1995000
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-base, designed for calculating scores of text pairs, suitable for text reordering and semantic search tasks.
Text Embedding English
R
ayushexel
30
0
Reranker ModernBERT Base Gooaq Bce
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-base for text re-ranking and semantic search tasks.
Text Embedding English
R
akr2002
16
1
Qwen2.5 14B YOYO V5
Apache-2.0
The fifth-generation Qwen2.5-YOYO model integrates features from multiple advanced models, optimizes the model merging formula, and supports a context length of 1 million tokens.
Large Language Model Supports Multiple Languages
Q
YOYO-AI
33
3
Reranker ModernBERT Base Gooaq Bce
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-base for text re-ranking and semantic search tasks.
Text Embedding English
R
tomaarsen
483
2
Reranker Msmarco ModernBERT Base Lambdaloss
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-base, designed for calculating scores of text pairs, suitable for text re-ranking and semantic search tasks.
Text Embedding English
R
tomaarsen
89
4
EXAONE Deep 2.4B AWQ
Other
The EXAONE Deep series models excel in reasoning tasks such as mathematics and programming. This model is the 4-bit AWQ quantized version with 2.4 billion parameters
Large Language Model Transformers Supports Multiple Languages
E
LGAI-EXAONE
751
16
Kyro N1.1 7B
Other
Kyro-n1.1 is an enhanced version of Kyro-n1, built upon Qwen2.5-7B-Instruct, with optimized reasoning capabilities, comprehension, and response accuracy.
Large Language Model Transformers Supports Multiple Languages
K
open-neo
470
4
Duo
Apache-2.0
DUO is a pretrained model based on the Transformers library, focusing on masked language modeling tasks and suitable for the field of natural language processing.
Large Language Model Transformers English
D
s-sahoo
212
1
ARWKV R1 1B5
Apache-2.0
ARWKV-R1-1B5 is an early preview version of a 7-billion-parameter model based on RNN, trained through three-stage knowledge distillation from DeepSeek-R1-Distill-Qwen-1.5B, with a context length of 2k.
Large Language Model Transformers Supports Multiple Languages
A
RWKV-Red-Team
164
4
Summllama3.2 3B Q4 0 GGUF
This is a GGUF format model converted from DISLab/SummLlama3.2-3B, primarily used for text summarization tasks.
Large Language Model
S
fernandoruiz
17
0
Lightnovel Translate Qwen2.5 32B GGUF
Apache-2.0
A large language model fine-tuned for Japanese-to-English translation of light novels and web novels, supporting long-text translation up to 32K tokens
Machine Translation Supports Multiple Languages
L
thefrigidliquidation
45
1
Gguf Jina Reranker V1 Tiny En
Apache-2.0
A model specifically designed for ultra-fast reranking, based on the JinaBERT architecture, supporting long text sequence processing (up to 8,192 tokens).
Text Embedding English
G
Felladrin
3,831
1
Modernbert Embed Base Legal MRL
Apache-2.0
A legal domain sentence embedding model fine-tuned based on ModernBERT, supporting multi-level dimensional output, suitable for legal text similarity calculation and information retrieval tasks.
Text Embedding English
M
AdamLucek
40
4
Bge M3 Distill 8l
An 8-layer embedding model distilled from BAAI/bge-m3, achieving 2.5x speed improvement while maintaining retrieval performance
Text Embedding
B
altaidevorg
249
7
Eagle2 2B
Eagle2 is a high-performance vision-language model family introduced by NVIDIA, focusing on enhancing the performance of open-source vision-language models through data strategies and training approaches. Eagle2-2B is the lightweight model in this series, achieving outstanding efficiency and speed while maintaining robust performance.
Text-to-Image Transformers Other
E
nvidia
667
21
Eagle2 9B
Eagle2-9B is the latest Vision-Language Model (VLM) released by NVIDIA, achieving a perfect balance between performance and inference speed. It is built on the Qwen2.5-7B-Instruct language model and the Siglip+ConvNext vision model, supporting multilingual and multimodal tasks.
Image-to-Text Transformers Other
E
nvidia
944
52
Openbuddy Deepseek 10b V17.1 4k GGUF
Other
GGUF format model file of OpenBuddy/openbuddy-deepseek-10b-v17.1-4k, supporting multiple languages, suitable for text generation tasks.
Large Language Model Supports Multiple Languages
O
tensorblock
81
1
Jina Embeddings V2 Base Code GGUF
Apache-2.0
Jina Embeddings V2 Base Code is a transformer-based English sentence embedding model, specializing in feature extraction and similarity computation for code-related texts.
Text Embedding English
J
second-state
315
1
Vapor V2 7B
Apache-2.0
A large language model fine-tuned on multilingual datasets based on Qwen/Qwen2.5-7B, supporting 13 language processing
Large Language Model Transformers
V
FourOhFour
60
4
BGE M3 Ko
Apache-2.0
A Korean-English bilingual sentence embedding model optimized based on BAAI/bge-m3, supporting semantic text similarity, information retrieval, and other tasks
Text Embedding Supports Multiple Languages
B
dragonkue
29.78k
44
Minicpm3 4B
Apache-2.0
MiniCPM3-4B is the third-generation model in the MiniCPM series, with overall performance surpassing Phi-3.5-mini-Instruct and GPT-3.5-Turbo-0125, comparable to several recent 7B~9B-scale models.
Large Language Model Transformers Supports Multiple Languages
M
openbmb
15.94k
414
Ko Reranker 8k
Apache-2.0
A text ranking model fine-tuned with Korean data based on BAAI/bge-reranker-v2-m3
Text Embedding Transformers Supports Multiple Languages
K
upskyy
14
11
Jais Family 30b 8k
Apache-2.0
The Jais series are large language models specialized in Arabic processing while possessing strong English capabilities, with variants ranging from 590 million to 70 billion parameters.
Large Language Model Supports Multiple Languages
J
inceptionai
48
9
Jais Family 6p7b Chat
Apache-2.0
The Jais family is a series of large bilingual language models specifically optimized for Arabic while maintaining strong English capabilities. This 6.7B parameter model is pre-trained from scratch and supports Arabic and English text generation.
Large Language Model Supports Multiple Languages
J
inceptionai
259
6
Bge M3 Spa Law Qa
Apache-2.0
A Spanish sentence embedding model fine-tuned based on BAAI/bge-m3, specifically optimized for the legal domain, suitable for semantic search and information retrieval tasks.
Text Embedding Spanish
B
littlejohn-ai
309
11
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase